Tailoring Boltzmann Machines to Netflix Data

نویسنده

  • Dustin Wehr
چکیده

If we let p(f) ∝ e, we obtain a probability distribution over assignments of binary states to the vertices of the machine. Let A(V ) denote the set of assigments of {0, 1} values to V , i.e. A(V ) = {a | a : V → {0, 1}}. A restricted boltzmann machine(RBM), introduced in [Smo86], is a boltzmann machine where the underlying graph is a complete bipartite graph. Let V = V1∪V2 be the partition of the vertex set. If we fix an assignment a1 ∈ A(V1) (respectively a2 ∈ A(V2)), we can compute the conditional distribution over A(V2) (resp. over A(V1)). Thus using Gibbs sampling we can sample from the joint distribution over A(V1) × A(V2). This leads to a tractable way of training an RBM to produce a desired probability distribution, see [HO06]. Salakhutdinov and Mnih obtained impressive results in the netflix competition by training an RBM to reproduce ratings data encoded as categorical data([SMH07]). We have attempted to improve upon their results by encorporating the ordinal nature of the data, and by explicitly considering the error function used by Netflix to evaluate the performance of a prediction model. In section 2 we discuss non-categorical representations of the ordinal ratings data. In section 3 we discuss the error function of the RBM, how it is different from the one specified by Netflix, and how we tried to overcome this. In section 4 we present the results of our experiments obtained by altering an RBM in the methods described. Finally in section 5 we offer our conclusions and future work.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

The BellKor Solution to the Netflix Grand Prize

This article describes part of our contribution to the “BellKor’s Pragmatic Chaos” final solution, which won the Netflix Grand Prize. The other portion of the contribution was created while working at AT&T with Robert Bell and Chris Volinsky, as reported in our 2008 Progress Prize report [3]. The final solution includes all the predictors described there. In this article we describe only the ne...

متن کامل

Combining Predictions for an accurate Recommender System

The application of ensemble learning to recommender systems is analyzed with the Netflix Prize dataset. We found that simple linear combination of predictions is not optimal in the sense of minimize the prediction RMSE. To predict ratings with collaborative filtering we use a set of predictions from different models (SVD, KNN, Restricted Boltzmann machine, Asymmetric Factor model, Global Effect...

متن کامل

Discriminative Restricted Boltzmann Machines are Universal Approximators for Discrete Data

This report proofs that discriminative Restricted Boltzmann Machines (RBMs) are universal approximators for discrete data by adapting existing universal approximation proofs for generative RBMs. Discriminative Restricted Boltzmann Machines are Universal Approximators for Discrete Data Laurens van der Maaten Pattern Recognition & Bioinformatics Laboratory Delft University of Technology

متن کامل

Sparse Group Restricted Boltzmann Machines

Since learning in Boltzmann machines is typically quite slow, there is a need to restrict connections within hidden layers. However, the resulting states of hidden units exhibit statistical dependencies. Based on this observation, we propose using l1/l2 regularization upon the activation probabilities of hidden units in restricted Boltzmann machines to capture the local dependencies among hidde...

متن کامل

Boltzmann machines and energy-based models

We review Boltzmann machines and energy-based models. A Boltzmann machine defines a probability distribution over binary-valued patterns. One can learn parameters of a Boltzmann machine via gradient based approaches in a way that log likelihood of data is increased. The gradient and Laplacian of a Boltzmann machine admit beautiful mathematical representations, although computing them is in gene...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2008